SdAE: Self-distillated Masked Autoencoder
نویسندگان
چکیده
With the development of generative-based self-supervised learning (SSL) approaches like BeiT and MAE, how to learn good representations by masking random patches input image reconstructing missing information has grown in concern. However, PeCo need a “pre-pretraining” stage produce discrete codebooks for masked representing. MAE does not require pre-training codebook process, but setting pixels as reconstruction targets may introduce an optimization gap between downstream tasks that quality always lead high descriptive capability model. Considering above issues, this paper, we propose simple Self-distillated AutoEncoder network, namely SdAE. SdAE consists student branch using encoder-decoder structure reconstruct information, teacher producing latent representation tokens. We also analyze build views from perspective bottleneck. After that, multi-fold strategy provide multiple with balanced boosting performance, which can reduce computational complexity. Our approach generalizes well: only 300 epochs pre-training, vanilla ViT-Base model achieves 84.1% fine-tuning accuracy on ImageNet-1 k classification, 48.6 mIOU ADE20K segmentation, 48.9 mAP COCO detection surpasses other methods considerable margin. Code is available at https://github.com/AbrahamYabo/SdAE .
منابع مشابه
MADE: Masked Autoencoder for Distribution Estimation
There has been a lot of recent interest in designing neural network models to estimate a distribution from a set of examples. We introduce a simple modification for autoencoder neural networks that yields powerful generative models. Our method masks the autoencoder’s parameters to respect autoregressive constraints: each input is reconstructed only from previous inputs in a given ordering. Cons...
متن کاملAutoencoder Trees
We discuss an autoencoder model in which the encoding and decoding functions are implemented by decision trees. We use the soft decision tree where internal nodes realize soft multivariate splits given by a gating function and the overall output is the average of all leaves weighted by the gating values on their path. The encoder tree takes the input and generates a lower dimensional representa...
متن کاملAutoEncoder by Forest
Auto-encoding is an important task which is typically realized by deep neural networks (DNNs) such as convolutional neural networks (CNN). In this paper, we propose EncoderForest (abbrv. eForest), the first tree ensemble based auto-encoder. We present a procedure for enabling forests to do backward reconstruction by utilizing the equivalent classes defined by decision paths of the trees, and de...
متن کاملCutting Recursive Autoencoder Trees
Deep Learning models enjoy considerable success in Natural Language Processing. While deep architectures produce useful representations that lead to improvements in various tasks, they are often difficult to interpret. This makes the analysis of learned structures particularly difficult. In this paper, we rely on empirical tests to see whether a particular structure makes sense. We present an a...
متن کاملAutoencoder for words
This paper presents a training method that encodes each word into a different vector in semantic space and its relation to low entropy coding. Elman network is employed in the method to process word sequences from literary works. The trained codes possess reduced entropy and are used in ranking, indexing, and categorizing literary works. A modification of the method to train the multi-vector fo...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Lecture Notes in Computer Science
سال: 2022
ISSN: ['1611-3349', '0302-9743']
DOI: https://doi.org/10.1007/978-3-031-20056-4_7